Goto

Collaborating Authors

 Oslo







OPEL: Optimal Transport Guided ProcedurE Learning

Neural Information Processing Systems

To this end, we propose to treat the video frames as samples from an unknown distribution, enabling us to frame their distance calculation as an optimal transport (OT) problem. Notably, the OT - based formulation allows us to relax the previously mentioned assumptions. To further improve performance, we enhance the OT formulation by introducing two regularization terms.


Representation Learning via Consistent Assignment of Views over Random Partitions

Neural Information Processing Systems

CARP learns prototypes in an end-to-end online fashion using gradient descent without additional non-differentiable modules to solve the cluster assignment problem. CARP optimizes a new pretext task based on random partitions of prototypes that regularizes the model and enforces consistency between views' assignments.



Aerial footage shows flooded cities as storms hit Spain

BBC News

Aerial footage showed the extend of floods in Spain after a series of storms hit the Iberian Peninsula. Storm Marta hit Spain on Saturday, bringing more rain to the region, as it was still recovering from Storm Leonardo. In Córdoba, drone footage showed flooded olive trees as Spanish farmers warned of the millions of euros worth of damage to crops following the torrential rains and high winds. In the country's southern region of Andalucia, over 11,000 people have been displaced. Nazar Daletskyi's relatives were told he had been killed in 2022, the first year of Russia's full-scale invasion.


Classifier Calibration at Scale: An Empirical Study of Model-Agnostic Post-Hoc Methods

Manokhin, Valery, Grønhaug, Daniel

arXiv.org Machine Learning

We study model-agnostic post-hoc calibration methods intended to improve probabilistic predictions in supervised binary classification on real i.i.d. tabular data, with particular emphasis on conformal and Venn-based approaches that provide distribution-free validity guarantees under exchangeability. We benchmark 21 widely used classifiers, including linear models, SVMs, tree ensembles (CatBoost, XGBoost, LightGBM), and modern tabular neural and foundation models, on binary tasks from the TabArena-v0.1 suite using randomized, stratified five-fold cross-validation with a held-out test fold. Five calibrators; Isotonic regression, Platt scaling, Beta calibration, Venn-Abers predictors, and Pearsonify are trained on a separate calibration split and applied to test predictions. Calibration is evaluated using proper scoring rules (log-loss and Brier score) and diagnostic measures (Spiegelhalter's Z, ECE, and ECI), alongside discrimination (AUC-ROC) and standard classification metrics. Across tasks and architectures, Venn-Abers predictors achieve the largest average reductions in log-loss, followed closely by Beta calibration, while Platt scaling exhibits weaker and less consistent effects. Beta calibration improves log-loss most frequently across tasks, whereas Venn-Abers displays fewer instances of extreme degradation and slightly more instances of extreme improvement. Importantly, we find that commonly used calibration procedures, most notably Platt scaling and isotonic regression, can systematically degrade proper scoring performance for strong modern tabular models. Overall classification performance is often preserved, but calibration effects vary substantially across datasets and architectures, and no method dominates uniformly. In expectation, all methods except Pearsonify slightly increase accuracy, but the effect is marginal, with the largest expected gain about 0.008%.